Goto

Collaborating Authors

 cubic term


Nonlinearity Enhanced Adaptive Activation Function

Yevick, David

arXiv.org Artificial Intelligence

While neural networks (NN) were first proposed in 1943 [1], initial implementations were restricted to networks with a small number of neurons and one or two layers[2], [3]. This limitation was eliminated through the backpropagation training algorithm[3], [4], [5] in conjunction with exponential improvements in computational performance. The resulting procedure generates a system model exclusively from experimental or simulated data and can accordingly be employed in a wide variety of scientific and engineering fields. In particular, a system, which typically can be characterized by a few coordinates and equations, is instead described by a large number of variables that interact nonlinearly. By optimizing a loss function, which may be further subject to physical constraints as in physics-informed machine 1 learning,[6] the parameters associated with the interactions are adjusted to approximate the data. The trained model then can predict the response of the system to unobserved input data. Although such an approach possesses significant advantages in terms of generality and simplicity, it lacks the precision and efficiency afforded by the solution of deterministic equations. Similarly, the large dimensionality of the representation obscures the underlying physics and mathematics. For complex systems, however, especially in the presence of stochastic noise or measurement inaccuracy, procedures based on numerical optimization can be effectively optimal.[7],


Augmented Electronic Ising Machine as an Effective SAT Solver

Sharma, Anshujit, Burns, Matthew, Hahn, Andrew, Huang, Michael

arXiv.org Artificial Intelligence

With the slowdown of improvement in conventional von Neumann systems, increasing attention is paid to novel paradigms such as Ising machines. They have very different approach to NP-complete optimization problems. Ising machines have shown great potential in solving binary optimization problems like MaxCut. In this paper, we present an analysis of these systems in satisfiability (SAT) problems. We demonstrate that, in the case of 3-SAT, a basic architecture fails to produce meaningful acceleration, thanks in no small part to the relentless progress made in conventional SAT solvers. Nevertheless, careful analysis attributes part of the failure to the lack of two important components: cubic interactions and efficient randomization heuristics. To overcome these limitations, we add proper architectural support for cubic interaction on a state-of-the-art Ising machine. More importantly, we propose a novel semantic-aware annealing schedule that makes the search-space navigation much more efficient than existing annealing heuristics. With experimental analyses, we show that such an Augmented Ising Machine for SAT (AIMS), outperforms state-of-the-art software-based, GPU-based and conventional hardware SAT solvers by orders of magnitude. We also demonstrate AIMS to be relatively robust against device variation and noise.


Efficient QUBO transformation for Higher Degree Pseudo Boolean Functions

Verma, Amit, Lewis, Mark, Kochenberger, Gary

arXiv.org Artificial Intelligence

Quadratic Unconstrained Binary Optimization (QUBO) is recognized as a unifying framework for modeling a wide range of problems. Problems can be solved with commercial solvers customized for solving QUBO and since QUBO have degree two, it is useful to have a method for transforming higher degree pseudo-Boolean problems to QUBO format. The standard transformation approach requires additional auxiliary variables supported by penalty terms for each higher degree term. This paper improves on the existing cubic-to-quadratic transformation approach by minimizing the number of additional variables as well as penalty coefficient. Extensive experimental testing on Max 3-SAT modeled as QUBO shows a near 100% reduction in the subproblem size used for minimization of the number of auxiliary variables.


Associative Memory in a Simple Model of Oscillating Cortex

Baird, Bill

Neural Information Processing Systems

A generic model of oscillating cortex, which assumes "minimal" sociative coupling justified by known anatomy, is shown to function as an as memory, using previously developed theory. The network has explicit excitatory neurons with local inhibitory interneuron feedback that forms a set of nonlinear oscillators coupled only by long range excitatofy connections. Using a local Hebb-like learning rule for primary and higher order synapses at the ends of the long range connections, the system learns to store the kinds of oscillation amplitude patterns observed in olfactory and visual cortex. This rule is derived from a more general "projection algorithm" for recurrent analog networks, that analytically guarantees content addressable memory storage of continuous periodic sequences - capacity: N /2 Fourier components for an N node network - no "spurious" attractors. 1 Introduction This is a sketch of recent results stemming from work which is discussed completely in [1, 2, 3]. Patterns of 40 to 80 hz oscillation have been observed in the large scale activity of olfactory cortex [4] and visual neocortex [5], and shown to predict the olfactory and visual pattern recognition responses of a trained animal. It thus appears that cortical computation in general may occur by dynamical interaction of resonant modes, as has been thought to be the case in the olfactory system.


Associative Memory in a Simple Model of Oscillating Cortex

Baird, Bill

Neural Information Processing Systems

A generic model of oscillating cortex, which assumes "minimal" coupling justified by known anatomy, is shown to function as an associative memory,using previously developed theory. The network has explicit excitatory neurons with local inhibitory interneuron feedback that forms a set of nonlinear oscillators coupled only by long range excitatofy connections. Using a local Hebb-like learning rule for primary and higher order synapses at the ends of the long range connections, the system learns to store the kinds of oscillation amplitudepatterns observed in olfactory and visual cortex. This rule is derived from a more general "projection algorithm" for recurrent analog networks, that analytically guarantees content addressable memory storage of continuous periodic sequences - capacity: N/2 Fourier components for an N node network - no "spurious" attractors. 1 Introduction This is a sketch of recent results stemming from work which is discussed completely in [1, 2, 3]. Patterns of 40 to 80 hz oscillation have been observed in the large scale activity of olfactory cortex [4] and visual neocortex [5], and shown to predict the olfactory and visual pattern recognition responses of a trained animal.